AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
ELECTRA-style Pretraining

# ELECTRA-style Pretraining

Intermediate Base
Apache-2.0
A Transformer model pre-trained on English corpus using ELECTRA-like self-supervised approach, suitable for sentence summarization tasks.
Large Language Model Transformers English
I
funnel-transformer
20
0
Deberta V3 Small
MIT
DeBERTa-v3 is an improved natural language understanding model developed by Microsoft, optimized through ELECTRA-style pretraining and gradient-disentangled embedding sharing technology to achieve efficient performance while maintaining a relatively small parameter count.
Large Language Model Transformers English
D
microsoft
189.23k
51
Deberta V3 Xsmall
MIT
DeBERTaV3 is an improved version of the DeBERTa model proposed by Microsoft, which enhances efficiency through ELECTRA-style gradient-disentangled embedding sharing pretraining method, demonstrating excellent performance in natural language understanding tasks.
Large Language Model Transformers English
D
microsoft
87.40k
43
Deberta V3 Base
MIT
DeBERTaV3 is an improved pre-trained language model based on DeBERTa, which enhances efficiency through gradient-disentangled embedding sharing in ELECTRA-style pretraining and excels in natural language understanding tasks.
Large Language Model English
D
microsoft
1.6M
316
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase